June 05, 2012 Volume 08 Issue 21
 

Designfax weekly eMagazine

Subscribe Today!
image of Designfax newsletter

Archives

View Archives

Partners

Manufacturing Center
Product Spotlight

Modern Applications News
Metalworking Ideas For
Today's Job Shops

Tooling and Production
Strategies for large
metalworking plants

Purdue researcher helps robots 'see' in 3D like humans

Zygmunt Pizlo, Purdue professor in the Dept. of Psychological Sciences, stands between postdoc research assistants Tadamasa Sawada and Yunfeng Li, as he adjusts the vision of a robot named Capek. The researchers move around the dance floor while the robot "watches." The goal is to simulate human visual perception in the robot. [Photo: Purdue Research Foundation]

 

 

 

 

Zygmunt Pizlo and his research team glide across a parquet dance floor - not in some club for a night on the town, but in his Purdue University Visual Perception Lab as part of critical research for a technology that is ready to be licensed and commercialized.

They're moving so a robot named Capek can "watch" them and conceptualize the research team's actions as members move around objects like desks and chairs. The goal is to simulate visual perception in the robot so it can "see" more like humans.

"Enabling robots and other machines to see the world in 3D like humans is one of the biggest challenges in robotics and artificial intelligence," said Pizlo, a professor in the Purdue Department of Psychological Sciences. "Research in the field of robotic vision has typically focused on recording and analyzing 2D images, but really it is about 3D visual perception - being able to understand the 3D scene in front of the robot so that it can decide what needs to be done with an object that is in its field of view. Should the robot walk around it? Pick it up?"

Pizlo has been working in the field of visual perception for 30 years.

"We believe there is a fundamental principle for human vision and that is we rely on a prior knowledge about a physical environment, so we're trying to program this knowledge of the physical environment into a robot's artificial intelligence," he said.

Pizlo said work in the lab is to develop a model of decision-making to mimic the human mind.

"We're developing a technology that will allow robotic machines to map their environment using two cameras for eyes," he said. "This process eliminates the need for additional range sensors currently used for robotic vision and reduces the time and complexity of robotic sight."

Pizlo and his research assistants demonstrate how a robot "sees" them moving across a dance floor in Pizlo's laboratory. [Photo: Purdue Research Foundation]

 

 

Postdoctoral research assistants Tadamasa Sawada and Yunfeng Li are working with Pizlo. Sawada said humans have the ability for cognitive functions that are computationally difficult, and it is a challenge to incorporate that ability into a robot.

"We quickly and easily perceive the physical world – a 3D shape and figure-ground organization," Sawada said. "Figure-ground organization is key to seeing an object in 3D instead of 2D."

Figure-ground perception is the tendency of a human's visual system to simplify a scene or photo into a main object and cognitively "move" everything else into the background.

"The question is how do we have a robot solve figure-ground organization like we do?" he said. "We do this by incorporating visual mechanisms and a prior knowledge about the physical environment into a robot."

Conventional robotic vision technology uses multiple cameras with laser range finders and other sensors to detect objects around them. While the current systems allow for basic object recognition, they do not replicate the 3D capabilities that are possible for humans, according to Pizlo.

"The key element in solving the puzzle of enabling 3D vision for robots is to realize that our visual system uses a prior knowledge about the physical world. We are implementing this knowledge, as well as the visual mechanisms (algorithms) in a robot," said Pizlo, who also is a professor of electrical and computer engineering. "The physical world is not completely random. Most natural objects are symmetrical, all objects have volume, gravity is always present, ground surfaces are approximately horizontal. There is enough prior knowledge in the human visual system that we automatically see everything in 3D. Combining a prior knowledge with visual information helps robots see the same way we do."

Enabling the human/robot connection will be a key factor in bringing robots into everyday life.

"Right now robots are used in a number of ways, including manufacturing, space research, agriculture and even cleaning our floors, but they can't bring us coffee in the morning," he said. "Until they can see like us, they can't truly interact with us. Once they can interact with us they can begin doing all types of tasks such as drive a car, help surgeons in hospitals, assist the elderly, provide sight for the blind, replace people in high-risk situations like making repairs in a nuclear plant and, yes, bring us coffee in the morning."

The patent-pending technology is available for licensing through Eric Lynch at 765-588-3477, eslynch@prf.org, in the Purdue Office of Technology Commercialization.

Funding for Pizlo' s research came from the National Science Foundation, U.S. Department of Defense, Air Force Office of Scientic Research, U.S. Department of Energy, and other sources.

Source: Purdue University

Published June 2012

Rate this article

[Purdue researcher helps robots 'see' in 3D like humans]

Very interesting, with information I can use
Interesting, with information I may use
Interesting, but not applicable to my operation
Not interesting or inaccurate

E-mail Address (required):

Comments:


Type the number:



Copyright © 2012 by Nelson Publishing, Inc. All rights reserved. Reproduction Prohibited.
View our terms of use and privacy policy